3 research outputs found
A Comprehensive Overview of Large Language Models
Large Language Models (LLMs) have shown excellent generalization capabilities
that have led to the development of numerous models. These models propose
various new architectures, tweaking existing architectures with refined
training strategies, increasing context length, using high-quality training
data, and increasing training time to outperform baselines. Analyzing new
developments is crucial for identifying changes that enhance training stability
and improve generalization in LLMs. This survey paper comprehensively analyses
the LLMs architectures and their categorization, training strategies, training
datasets, and performance evaluations and discusses future research directions.
Moreover, the paper also discusses the basic building blocks and concepts
behind LLMs, followed by a complete overview of LLMs, including their important
features and functions. Finally, the paper summarizes significant findings from
LLM research and consolidates essential architectural and training strategies
for developing advanced LLMs. Given the continuous advancements in LLMs, we
intend to regularly update this paper by incorporating new sections and
featuring the latest LLM models
Energy level spoofing attacks and countermeasures in blockchain-enabled IoT
The Internet of Things (IoT) ecosystem is witnessing widespread deployments for emerging applications in diverse domains such as remote sensing, smart homes, and industry 4.0. There is also a growing need to secure such deployments against malicious IoT devices to sustain normal network operations. Since the IoT deployments encompass geographically distributed nodes, blockchain technology, which inherently offers distributed trust in such scenarios, is gaining popularity in providing a secure and trusted IoT deployment. In this paper, we present a use case in which an IoT deployment is retrofitted with a blockchain. The use of blockchain prevents malicious nodes from falsifying information about their energy levels. We first present attack scenarios where IoT nodes can spoof energy while joining or being a part of the network. We then build a defense strategy and evaluate its performance under various attack scenarios. Our results indicate that the IoT deployment is robust under the proposed defense strategy which can detect if a node is spoofing its energy levels over 75% of the time
Survey: Image Mixing and Deleting for Data Augmentation
Data augmentation has been widely used to improve deep nerual networks
performance. Numerous approaches are suggested, for example, dropout,
regularization and image augmentation, to avoid over-ftting and enhancing
generalization of neural networks. One of the sub-area within data augmentation
is image mixing and deleting. This specific type of augmentation either mixes
two images or delete image regions to hide or make certain characteristics of
images confusing for the network to force it to emphasize on overall structure
of object in image. The model trained with this approach has shown to perform
and generalize well as compared to one trained without imgage mixing or
deleting. Additional benefit achieved with this method of training is
robustness against image corruptions. Due to its low compute cost and success
in recent past, many techniques of image mixing and deleting are proposed. This
paper provides detailed review on these devised approaches, dividing
augmentation strategies in three main categories cut and delete, cut and mix
and mixup. The second part of paper emprically evaluates these approaches for
image classification, finegrained image recognition and object detection where
it is shown that this category of data augmentation improves the overall
performance for deep neural networks